5 research outputs found
SizeNet: Weakly Supervised Learning of Visual Size and Fit in Fashion Images
Finding clothes that fit is a hot topic in the e-commerce fashion industry.
Most approaches addressing this problem are based on statistical methods
relying on historical data of articles purchased and returned to the store.
Such approaches suffer from the cold start problem for the thousands of
articles appearing on the shopping platforms every day, for which no prior
purchase history is available. We propose to employ visual data to infer size
and fit characteristics of fashion articles. We introduce SizeNet, a
weakly-supervised teacher-student training framework that leverages the power
of statistical models combined with the rich visual information from article
images to learn visual cues for size and fit characteristics, capable of
tackling the challenging cold start problem. Detailed experiments are performed
on thousands of textile garments, including dresses, trousers, knitwear, tops,
etc. from hundreds of different brands.Comment: IEEE Conference on Computer Vision and Pattern Recognition Workshop
(CVPRW) 2019 Focus on Fashion and Subjective Search - Understanding
Subjective Attributes of Data (FFSS-USAD
Gaze Embeddings for Zero-Shot Image Classification
Zero-shot image classification using auxiliary information, such as
attributes describing discriminative object properties, requires time-consuming
annotation by domain experts. We instead propose a method that relies on human
gaze as auxiliary information, exploiting that even non-expert users have a
natural ability to judge class membership. We present a data collection
paradigm that involves a discrimination task to increase the information
content obtained from gaze data. Our method extracts discriminative descriptors
from the data and learns a compatibility function between image and gaze using
three novel gaze embeddings: Gaze Histograms (GH), Gaze Features with Grid
(GFG) and Gaze Features with Sequence (GFS). We introduce two new
gaze-annotated datasets for fine-grained image classification and show that
human gaze data is indeed class discriminative, provides a competitive
alternative to expert-annotated attributes, and outperforms other baselines for
zero-shot image classification
UNICON: A unified framework for behavior-based consumer segmentation in e-commerce
Data-driven personalization is a key practice in fashion e-commerce,
improving the way businesses serve their consumers needs with more relevant
content. While hyper-personalization offers highly targeted experiences to each
consumer, it requires a significant amount of private data to create an
individualized journey. To alleviate this, group-based personalization provides
a moderate level of personalization built on broader common preferences of a
consumer segment, while still being able to personalize the results. We
introduce UNICON, a unified deep learning consumer segmentation framework that
leverages rich consumer behavior data to learn long-term latent representations
and utilizes them to extract two pivotal types of segmentation catering various
personalization use-cases: lookalike, expanding a predefined target seed
segment with consumers of similar behavior, and data-driven, revealing
non-obvious consumer segments with similar affinities. We demonstrate through
extensive experimentation our framework effectiveness in fashion to identify
lookalike Designer audience and data-driven style segments. Furthermore, we
present experiments that showcase how segment information can be incorporated
in a hybrid recommender system combining hyper and group-based personalization
to exploit the advantages of both alternatives and provide improvements on
consumer experience
The Mental Image Revealed by Gaze Tracking
Humans involuntarily move their eyes when retrieving an
image from memory. This motion is often similar to actually
observing the image. We suggest to exploit this behavior as
a new modality in human computer interaction, using the
motion of the eyes as a descriptor of the image. Interaction
requires the user’s eyes to be tracked, but no voluntary physical activity.We perform a controlled experiment and develop matching techniques using machine learning to investigate if images can be discriminated based on the gaze patterns recorded while users merely recall an image. Our results indicate that image retrieval is possible with an accuracy significantly above chance. We also show that these results generalize to images not used during training of the classifier and extends to uncontrolled settings in a realistic scenario